Winner-Take-All Autoencoders

نویسندگان

  • Alireza Makhzani
  • Brendan J. Frey
چکیده

In this paper, we propose a winner-take-all method for learning hierarchical sparse representations in an unsupervised fashion. We first introduce fully-connected winner-take-all autoencoders which use mini-batch statistics to directly enforce a lifetime sparsity in the activations of the hidden units. We then propose the convolutional winner-take-all autoencoder which combines the benefits of convolutional architectures and autoencoders for learning shift-invariant sparse representations. We describe a way to train convolutional autoencoders layer by layer, where in addition to lifetime sparsity, a spatial sparsity within each feature map is achieved using winner-take-all activation functions. We will show that winner-take-all autoencoders can be used to to learn deep sparse representations from the MNIST, CIFAR-10, ImageNet, Street View House Numbers and Toronto Face datasets, and achieve competitive classification performance.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Winner-Take-All Method for Training Sparse Convolutional Autoencoders

We explore combining the benefits of convolutional architectures and autoencoders for learning deep representations in an unsupervised manner. A major challenge is to achieve appropriate sparsity among hidden variables, since neighbouring variables in each feature map tend to be highly correlated and a suppression mechanism is therefore needed. Previously, deconvolutional networks and convoluti...

متن کامل

Exploiting Spatio-Temporal Structure with Recurrent Winner-Take-All Networks

We propose a convolutional recurrent neural network, with Winner-Take-All dropout for high dimensional unsupervised feature learning in multi-dimensional time series. We apply the proposedmethod for object recognition with temporal context in videos and obtain better results than comparable methods in the literature, including the Deep Predictive Coding Networks previously proposed by Chalasani...

متن کامل

Neural Computation with Winner-Take-All as the Only Nonlinear Operation

Everybody “knows” that neural networks need more than a single layer of nonlinear units to compute interesting functions. We show that this is false if one employs winner-take-all as nonlinear unit: Any boolean function can be computed by a single -winner-takeall unit applied to weighted sums of the input variables. Any continuous function can be approximated arbitrarily well by a single soft w...

متن کامل

Spectra of winner-take-all stochastic neural networks

In Piekniewski & Schreiber (2008) we have developed a simple mathematical model for information flow structure in a class of recurrent neural networks and shown that its asymptotic behaviour is scale-free and admits a description in terms of the so-called winner-take-all dynamics. In the present paper we establish a limit theorem for spectra of the spike-flow graphs induced by the winner-take-a...

متن کامل

A Fast Winner-Take-All Neural Networks With the Dynamic Ratio

In this paper, we propose a fast winner-take-all (WTA) neural network. The fast winner-take-all neural network with the dynamic ratio in mutual-inhibition is developed from the general mean-based neural network (GEMNET), which adopts the mean of the active neurons as the threshold of mutual inhibition. Furthermore, the other winner-take-all neural network enhances the convergence speed to becom...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2015